# Low-Resource Inference

Motif 2.6B
Other
Motif 2.6B is a language model with 2.6 billion parameters, trained from scratch on AMD Instinct™ MI250 GPUs, aiming to build AI that aligns with human values, is useful, and reliable.
Large Language Model Safetensors Supports Multiple Languages
M
Motif-Technologies
1,470
29
Phantom Wan 1.3B GGUF
Apache-2.0
This is a direct GGUF conversion version of the bytedance-research/Phantom model, usable in ComfyUI with the ComfyUI-GGUF custom node.
Text-to-Video English
P
QuantStack
174
0
Llava 1.5 7b Hf Q4 K M GGUF
This model is a GGUF format conversion of llava-hf/llava-1.5-7b-hf, supporting image-to-text generation tasks.
Image-to-Text English
L
Marwan02
30
1
Seed Coder 8B Reasoning Bf16 Q6 K GGUF
MIT
This is a GGUF format model converted from ByteDance-Seed/Seed-Coder-8B-Reasoning-bf16, suitable for code generation and reasoning tasks.
Large Language Model
S
GrimsenClory
100
1
Qwen3 8B GGUF
Apache-2.0
Qwen3 is the latest iteration of the Tongyi Qianwen series of large language models, offering a complete suite of dense models and Mixture of Experts (MoE) models. Based on large-scale training, Qwen3 achieves breakthrough progress in reasoning capabilities, instruction following, agent functionalities, and multilingual support.
Large Language Model
Q
Qwen
4,474
8
Ring Lite Linear Preview
MIT
The Linglong Linear Preview is a hybrid linear sparse large language model open-sourced by InclusionAI, with a total of 17.1B parameters and 3.0B activated parameters. This model implements long-text reasoning based on a hybrid linear attention mechanism, achieving near-linear computational complexity and near-constant space complexity during inference.
Large Language Model Supports Multiple Languages
R
inclusionAI
25
8
Orpheus 3b Kaya Q4 K M.gguf
Apache-2.0
A fine-tuned text-to-speech model based on Canopy Labs' pre-trained model, quantized for efficient inference
Speech Synthesis Supports Multiple Languages
O
lex-au
98
0
Orpheus 3b Kaya FP16.gguf
Apache-2.0
A text-to-speech (TTS) model fine-tuned from Canopy Labs' pre-trained model, quantized in GGUF FP16 format for efficient inference
Speech Synthesis Supports Multiple Languages
O
lex-au
45
0
Phi 4 Mini Instruct 8da4w
MIT
Phi-4-mini is a quantized language model developed by the PyTorch team, featuring 8-bit embeddings, 8-bit dynamic activations, and a 4-bit weight linear layer (8da4w) quantization scheme, making it suitable for mobile deployment.
Large Language Model Transformers Other
P
pytorch
780
1
Qwen2.5 VL 7B Instruct Q8 0 GGUF
Apache-2.0
This model is a GGUF-format conversion of Qwen2.5-VL-7B-Instruct, supporting multimodal tasks and applicable to image and text interaction processing.
Text-to-Image English
Q
cxtb
72
1
Qwen2.5 VL 7B Instruct Q4 K M GGUF
Apache-2.0
This is the GGUF quantized version of the Qwen2.5-VL-7B-Instruct model, suitable for multimodal tasks and supports both image and text inputs.
Image-to-Text English
Q
PatataAliena
69
1
Fibonacci 2 14B
MIT
A large language model based on the Phi 4 architecture, with 14 billion parameters, optimized for natural language processing and text dialogue tasks.
Large Language Model Supports Multiple Languages
F
fibonacciai
97
13
Mlabonne Gemma 3 4b It Abliterated GGUF
This is a quantized version based on the mlabonne/gemma-3-4b-it-abliterated model, using llama.cpp for imatrix quantization, suitable for image-text-to-text tasks.
Image-to-Text
M
bartowski
9,164
8
RWKV7 Goose Pile 168M HF
Apache-2.0
RWKV-7 model using Flash Linear Attention format, trained on the Pile dataset, supporting English text generation tasks.
Large Language Model Transformers English
R
RWKV
57
2
Open R1 OlympicCoder 32B GGUF
Apache-2.0
Quantized version of OlympicCoder-32B, based on llama.cpp's imatrix quantization method, suitable for code generation tasks.
Large Language Model English
O
bartowski
12.60k
12
Gemmax2 28 2B Gguf
Apache-2.0
The GemmaX2-28-2B GGUF quantized model is a series of quantized variants based on GemmaX2-28-2B-v0.1, designed for multilingual machine translation and supports 28 languages.
Machine Translation Supports Multiple Languages
G
Tonic
258
5
Ozone Ai 0x Lite GGUF
Apache-2.0
Quantized version based on ozone-ai/0x-lite model, supporting Chinese and English text generation tasks, using llama.cpp for imatrix quantization, offering multiple quantization options to adapt to different hardware requirements.
Large Language Model Supports Multiple Languages
O
bartowski
220
2
Llama 3.1 8B Instuct Uz Q4 K M GGUF
This is an 8B-parameter model based on the Llama-3.1 architecture, specifically optimized for Uzbek and English, supporting tasks such as text generation, summarization, translation, and Q&A.
Large Language Model Supports Multiple Languages
L
azimjon
19
0
Rwkv7 1.5B World
Apache-2.0
The RWKV-7 model adopts a flash linear attention architecture and supports multilingual text generation tasks.
Large Language Model Transformers Supports Multiple Languages
R
fla-hub
632
9
Qwen2 VL 7B Captioner Relaxed Q4 K M GGUF
Apache-2.0
This is a GGUF format model converted from the Qwen2-VL-7B-Captioner-Relaxed model, specifically designed for image-to-text tasks.
Image-to-Text English
Q
alecccdd
88
1
Senecallm X Qwen2.5 7B CyberSecurity Q8 0 GGUF
MIT
SenecaLLM is a large language model fine-tuned based on Qwen2.5-Coder-7B-Instruct, specializing in cybersecurity-related tasks.
Large Language Model English
S
AlicanKiraz0
84
8
Meta Llama 3.1 8B Instruct GGUF
Other
Llama 3.1 8B Instruct is a large language model released by Meta. It has been fixed and optimized for a specific version, improving performance and compatibility.
Large Language Model English
M
aniljava
158
8
QQQ Llama 3 8b G128
MIT
This is a version of the Llama-3-8b model quantized to INT4, using the QQQ quantization technique with a group size of 128 and optimized for hardware.
Large Language Model Transformers
Q
HandH1998
1,708
2
Llava Llama 3 8b V1 1 Q4 K M GGUF
This model is a GGUF format conversion based on xtuner/llava-llama-3-8b-v1_1, supporting multimodal interaction between images and text.
Image-to-Text
L
RaincloudAi
51
1
Eris PrimeV3 Vision 7B
Other
Eris Prime V2 is a 7B-parameter multimodal language model with vision capabilities, requiring Koboldcpp for operation.
Text-to-Image
E
ChaoticNeutrals
118
8
Mixtral 8x7B V0.1 GGUF
Apache-2.0
GGUF quantized version of Mixtral-8x7B-v0.1, supporting multiple bit quantization, suitable for text generation tasks.
Large Language Model Supports Multiple Languages
M
MaziyarPanahi
128
1
Deepseek Coder 1.3b Instruct GPTQ
Other
The GPTQ quantized version of Deepseek Coder 1.3B Instruct, offering multiple quantization parameter options, suitable for code generation and computer science-related tasks.
Large Language Model Transformers
D
TheBloke
653
7
Hallucination Evaluation Model
Apache-2.0
HHEM-2.1-Open is a hallucination detection model developed by Vectara, designed to evaluate the consistency between content generated by large language models and given evidence.
Large Language Model Transformers English
H
vectara
229.46k
280
Llava V1.5 13B GPTQ
Llava v1.5 13B is a multimodal model developed by Haotian Liu, combining visual and language capabilities to understand and generate content based on images and text.
Text-to-Image Transformers
L
TheBloke
131
37
Mistral Trismegistus 7B
Apache-2.0
The Mistral Trismegistus 7B Model is a professional-grade large language model specializing in esotericism, metaphysics, and spirituality. It is based on the Mistral-7B architecture and fine-tuned with synthetic data generated by GPT-4.
Large Language Model Transformers English
M
teknium
54
218
Codellama 7B GGUF
CodeLlama 7B is a 7B-parameter code generation and comprehension model developed by Meta, optimized based on the Llama 2 architecture and focused on programming tasks.
Large Language Model Other
C
TheBloke
10.80k
121
Mythomax L2 13B GPTQ
Other
MythoMax L2 13B is a large language model developed by Gryphe, based on the Llama 2 architecture, focusing on role-playing and creative text generation.
Large Language Model Transformers English
M
TheBloke
5,324
204
Replit Code V1 3b
A 2.7 billion parameter code generation model developed by Replit, supporting 20 programming languages
Large Language Model Transformers Other
R
replit
605
733
Cadet Tiny
Openrail
Cadet-Tiny is an ultra-compact dialogue model trained on the SODA dataset, specifically designed for edge device inference, with a size of only about 2% of the Cosmo-3B model.
Dialogue System Transformers English
C
ToddGoldfarb
2,691
6
Bert Base Uncased Squadv1 X2.32 F86.6 D15 Hybrid V1
MIT
A QA model fine-tuned on SQuAD v1 based on BERT-base uncased, with 66% of linear layer weights pruned via nn_pruning library, achieving 2.32x inference speedup
Question Answering System Transformers English
B
madlag
16
0
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase